Goto

Collaborating Authors

 global distribution






C${}^2$Prompt: Class-aware Client Knowledge Interaction for Federated Continual Learning

Xu, Kunlun, Feng, Yibo, Li, Jiangmeng, Qi, Yongsheng, Zhou, Jiahuan

arXiv.org Artificial Intelligence

Federated continual learning (FCL) tackles scenarios of learning from continuously emerging task data across distributed clients, where the key challenge lies in addressing both temporal forgetting over time and spatial forgetting simultaneously. Recently, prompt-based FCL methods have shown advanced performance through task-wise prompt communication.In this study, we underscore that the existing prompt-based FCL methods are prone to class-wise knowledge coherence between prompts across clients. The class-wise knowledge coherence includes two aspects: (1) intra-class distribution gap across clients, which degrades the learned semantics across prompts, (2) inter-prompt class-wise relevance, which highlights cross-class knowledge confusion. During prompt communication, insufficient class-wise coherence exacerbates knowledge conflicts among new prompts and induces interference with old prompts, intensifying both spatial and temporal forgetting. To address these issues, we propose a novel Class-aware Client Knowledge Interaction (C${}^2$Prompt) method that explicitly enhances class-wise knowledge coherence during prompt communication. Specifically, a local class distribution compensation mechanism (LCDC) is introduced to reduce intra-class distribution disparities across clients, thereby reinforcing intra-class knowledge consistency. Additionally, a class-aware prompt aggregation scheme (CPA) is designed to alleviate inter-class knowledge confusion by selectively strengthening class-relevant knowledge aggregation. Extensive experiments on multiple FCL benchmarks demonstrate that C${}^2$Prompt achieves state-of-the-art performance. Our source code is available at https://github.com/zhoujiahuan1991/NeurIPS2025-C2Prompt


Calibrating Biased Distribution in VFM-derived Latent Space via Cross-Domain Geometric Consistency

Ma, Yanbiao, Dai, Wei, Liu, Bowei, Chen, Jiayi, Huang, Wenke, Wan, Guancheng, Lu, Zhiwu, Yan, Junchi

arXiv.org Artificial Intelligence

Abstract--Despite the fast progress of deep learning, one standing challenge is the gap of the observed training samples and the underlying true distribution. There are multiple reasons for the causing of this gap e.g. In the era of foundation models, we show that when leveraging the off-the-shelf (vision) foundation models (e.g., CLIP, DINOv2) for feature extraction, the geometric shapes of the resulting feature distributions exhibit remarkable transferability across domains and datasets. T o verify its practical usefulness, we embody our geometric knowledge-guided distribution calibration framework in two popular and challenging settings: federated learning and long-tailed recognition. In the federated setting, we devise a technique of acquiring the global geometric shape under privacy constraints, then leverage this knowledge to generate new samples for clients, in the aim of bridging the gap between local and global observations. In long-tailed learning, it utilizes the geometric knowledge transferred from sample-rich categories to recover the true distribution for sample-scarce tail classes. Comprehensive experiments show that our proposed geometric knowledge-guided distribution calibration effectively overcomes information deficits caused by data heterogeneity and sample imbalance, with boosted performance across benchmarks. It is often the case that the training data relied upon by models is often only a local [6], sparse [7], and biased observation [8] of the underlying ideal global data distribution. This distribution missing phenomenon manifests in various forms: in federated learning, it appears as label skew and domain skew due to data silos among clients [9], [10], [11], causing a severe misalignment between local data distributions and the global ideal distribution, thereby leading to divergent or even conflicting local optimization directions [12], [13], [14]. In long-tailed recognition, it is characterized by the extreme scarcity of samples in tail classes, preventing the model from capturing the true and complete shape of their distributions [15], [16]. Despite the differing scenarios, the essence is highly unified--models learn from incomplete information, lacking a comprehensive understanding of the overall structure of the real world. Conventional solutions, such as weighting loss functions [7], [17], [18], designing complex regularization terms [9], [14], [19], or aggregation strategies [20], [21], [22], primarily focus on post-hoc compensation at the optimization level. Y anbiao Ma and Zhiwu Lu are with the Gaoling School of Artificial Intelligence, Renmin University of China. Bowen Liu is with T singhua University.


Geometric Knowledge-Guided Localized Global Distribution Alignment for Federated Learning

Ma, Yanbiao, Dai, Wei, Huang, Wenke, Chen, Jiayi

arXiv.org Artificial Intelligence

Data heterogeneity in federated learning, characterized by a significant misalignment between local and global distributions, leads to divergent local optimization directions and hinders global model training. Existing studies mainly focus on optimizing local updates or global aggregation, but these indirect approaches demonstrate instability when handling highly heterogeneous data distributions, especially in scenarios where label skew and domain skew coexist. To address this, we propose a geometry-guided data generation method that centers on simulating the global embedding distribution locally. We first introduce the concept of the geometric shape of an embedding distribution and then address the challenge of obtaining global geometric shapes under privacy constraints. Subsequently, we propose GGEUR, which leverages global geometric shapes to guide the generation of new samples, enabling a closer approximation to the ideal global distribution. In single-domain scenarios, we augment samples based on global geometric shapes to enhance model generalization; in multi-domain scenarios, we further employ class prototypes to simulate the global distribution across domains. Extensive experimental results demonstrate that our method significantly enhances the performance of existing approaches in handling highly heterogeneous data, including scenarios with label skew, domain skew, and their coexistence. Code published at: https://github.com/WeiDai-David/2025CVPR_GGEUR


AFed: Algorithmic Fair Federated Learning

Chen, Huiqiang, Zhu, Tianqing, Zhou, Wanlei, Zhao, Wei

arXiv.org Artificial Intelligence

Federated Learning (FL) has gained significant attention as it facilitates collaborative machine learning among multiple clients without centralizing their data on a server. FL ensures the privacy of participating clients by locally storing their data, which creates new challenges in fairness. Traditional debiasing methods assume centralized access to sensitive information, rendering them impractical for the FL setting. Additionally, FL is more susceptible to fairness issues than centralized machine learning due to the diverse client data sources that may be associated with group information. Therefore, training a fair model in FL without access to client local data is important and challenging. This paper presents AFed, a straightforward yet effective framework for promoting group fairness in FL. The core idea is to circumvent restricted data access by learning the global data distribution. This paper proposes two approaches: AFed-G, which uses a conditional generator trained on the server side, and AFed-GAN, which improves upon AFed-G by training a conditional GAN on the client side. We augment the client data with the generated samples to help remove bias. Our theoretical analysis justifies the proposed methods, and empirical results on multiple real-world datasets demonstrate a substantial improvement in AFed over several baselines.


On Homomorphic Encryption Based Strategies for Class Imbalance in Federated Learning

Guleria, Arpit, Harshan, J., Prasad, Ranjitha, Bharath, B. N.

arXiv.org Artificial Intelligence

Class imbalance in training datasets can lead to bias and poor generalization in machine learning models. While pre-processing of training datasets can efficiently address both these issues in centralized learning environments, it is challenging to detect and address these issues in a distributed learning environment such as federated learning. In this paper, we propose FLICKER, a privacy preserving framework to address issues related to global class imbalance in federated learning. At the heart of our contribution lies the popular CKKS homomorphic encryption scheme, which is used by the clients to privately share their data attributes, and subsequently balance their datasets before implementing the FL scheme. Extensive experimental results show that our proposed method significantly improves the FL accuracy numbers when used along with popular datasets and relevant baselines.


FOOGD: Federated Collaboration for Both Out-of-distribution Generalization and Detection

Liao, Xinting, Liu, Weiming, Zhou, Pengyang, Yu, Fengyuan, Xu, Jiahe, Wang, Jun, Wang, Wenjie, Chen, Chaochao, Zheng, Xiaolin

arXiv.org Artificial Intelligence

Federated learning (FL) is a promising machine learning paradigm that collaborates with client models to capture global knowledge. However, deploying FL models in real-world scenarios remains unreliable due to the coexistence of in-distribution data and unexpected out-of-distribution (OOD) data, such as covariate-shift and semantic-shift data. Current FL researches typically address either covariate-shift data through OOD generalization or semantic-shift data via OOD detection, overlooking the simultaneous occurrence of various OOD shifts. In this work, we propose FOOGD, a method that estimates the probability density of each client and obtains reliable global distribution as guidance for the subsequent FL process. Firstly, SM3D in FOOGD estimates score model for arbitrary distributions without prior constraints, and detects semantic-shift data powerfully. Then SAG in FOOGD provides invariant yet diverse knowledge for both local covariate-shift generalization and client performance generalization. In empirical validations, FOOGD significantly enjoys three main advantages: (1) reliably estimating non-normalized decentralized distributions, (2) detecting semantic shift data via score values, and (3) generalizing to covariate-shift data by regularizing feature extractor. The prejoct is open in https://github.com/XeniaLLL/FOOGD-main.git.